extracting finite state machine
Extracting Finite State Machines from Transformers
Fueled by the popularity of the transformer architecture in deep learning, several works have investigated what formal languages a transformer can learn. Nonetheless, existing results remain hard to compare and a fine-grained understanding of the trainability of transformers on regular languages is still lacking. We investigate transformers trained on regular languages from a mechanistic interpretability perspective. Using an extension of the $L^*$ algorithm, we extract Moore machines from transformers. We empirically find tighter lower bounds on the trainability of transformers, when a finite number of symbols determine the state. Additionally, our mechanistic insight allows us to characterise the regular languages a one-layer transformer can learn with good length generalisation. However, we also identify failure cases where the determining symbols get misrecognised due to saturation of the attention mechanism.
- North America > United States (0.14)
- Europe > Austria > Vienna (0.14)
- Europe > Belgium > Flanders > Flemish Brabant > Leuven (0.04)
- (4 more...)
Fool's Gold: Extracting Finite State Machines from Recurrent Network Dynamics
Several recurrent networks have been proposed as representations for the task of formal language learning. After training a recurrent network rec(cid:173) ognize a formal language or predict the next symbol of a sequence, the next logical step is to understand the information processing carried out by the network. Some researchers have begun to extracting finite state machines from the internal state trajectories of their recurrent networks. This paper describes how sensitivity to initial conditions and discrete measurements can trick these extraction methods to return illusory finite state descriptions.
Fool's Gold: Extracting Finite State Machines from Recurrent Network Dynamics
Several recurrent networks have been proposed as representations for the task of formal language learning. After training a recurrent network recognize a formal language or predict the next symbol of a sequence, the next logical step is to understand the information processing carried out by the network. Some researchers have begun to extracting finite state machines from the internal state trajectories of their recurrent networks. This paper describes how sensitivity to initial conditions and discrete measurements can trick these extraction methods to return illusory finite state descriptions.
- North America > United States > Massachusetts > Middlesex County > Cambridge (0.14)
- North America > United States > Ohio > Franklin County > Columbus (0.04)
- North America > United States > California > San Francisco County > San Francisco (0.04)
- (2 more...)
Fool's Gold: Extracting Finite State Machines from Recurrent Network Dynamics
Several recurrent networks have been proposed as representations for the task of formal language learning. After training a recurrent network recognize aformal language or predict the next symbol of a sequence, the next logical step is to understand the information processing carried out by the network. Some researchers have begun to extracting finite state machines from the internal state trajectories of their recurrent networks. This paper describes how sensitivity to initial conditions and discrete measurements can trick these extraction methods to return illusory finite state descriptions.
- North America > United States > Massachusetts > Middlesex County > Cambridge (0.14)
- North America > United States > Ohio > Franklin County > Columbus (0.04)
- North America > United States > California > San Francisco County > San Francisco (0.04)
- (2 more...)
Fool's Gold: Extracting Finite State Machines from Recurrent Network Dynamics
Several recurrent networks have been proposed as representations for the task of formal language learning. After training a recurrent network recognize a formal language or predict the next symbol of a sequence, the next logical step is to understand the information processing carried out by the network. Some researchers have begun to extracting finite state machines from the internal state trajectories of their recurrent networks. This paper describes how sensitivity to initial conditions and discrete measurements can trick these extraction methods to return illusory finite state descriptions.
- North America > United States > Massachusetts > Middlesex County > Cambridge (0.14)
- North America > United States > Ohio > Franklin County > Columbus (0.04)
- North America > United States > California > San Francisco County > San Francisco (0.04)
- (2 more...)